ParC-Net: Position Aware Circular Convolution with Merits from ConvNets and Transformer
نویسندگان
چکیده
Recently, vision transformers started to show impressive results which outperform large convolution based models significantly. However, in the area of small for mobile or resource constrained devices, ConvNet still has its own advantages both performance and model complexity. We propose ParC-Net, a pure backbone that further strengthens these by fusing merits into ConvNets. Specifically, we p osition ware ci r cular c onvolution (ParC), light-weight op boasts global receptive field while producing location sensitive features as local convolutions. combine ParCs squeeze-excitation ops form meta-former like block, attention mechanism transformers. The aforementioned block can be used plug-and-play manner replace relevant blocks ConvNets Experiment proposed ParC-Net achieves better than popular transformer common tasks datasets, having fewer parameters faster inference speed. For classification on ImageNet-1k, 78.6% top-1 accuracy with about 5.0 million parameters, saving 11% 13% computational cost but gaining 0.2% higher 23% speed (on ARM Rockchip RK3288) compared MobileViT, uses only 0.5 $$\times $$ 2.7% DeIT. On MS-COCO object detection PASCAL VOC segmentation tasks, also shows performance. Source code is available at https://github.com/hkzhang91/ParC-Net .
منابع مشابه
ParC#: Parallel Computing with C# in .Net
This paper describes experiments with the development of a parallel computing platform on top of a compatible C# implementation: the Mono project. This implementation has the advantage of running on both Windows and UNIX platforms and has reached a stable state. This paper presents performance results obtained and compares these results with implementations in Java/RMI. The results show that th...
متن کاملConvolution Aware Initialization
Initialization of parameters in deep neural networks has been shown to have a big impact on the performance of the networks (Mishkin & Matas, 2015). The initialization scheme devised by He et al, allowed convolution activations to carry a constrained mean which allowed deep networks to be trained effectively (He et al., 2015a). Orthogonal initializations and more generally orthogonal matrices i...
متن کاملDiscrete Fresnel Transform and Its Circular Convolution
Abstract—Discrete trigonometric transformations, such as the discrete Fourier and cosine/sine transforms, are important in a variety of applications due to their useful properties. For example, one well-known property is the convolution theorem for Fourier transform. In this letter, we derive a discrete Fresnel transform (DFnT) from the infinitely periodic optical gratings, as a linear trigon...
متن کاملPosition-aware activity recognition with wearable devices
Reliable human activity recognition with wearable devices enables the development of human-centric pervasive applications.We aim to develop a robust wearable-based activity recognition system for real life situations where the device position is up to the user or where a user is unable to collect initial training data. Consequently, in this work we focus on the problem of recognizing the on-bod...
متن کاملFast Reading Comprehension with ConvNets
State-of-the-art deep reading comprehension models are dominated by recurrent neural nets. Their sequential nature is a natural fit for language, but it also precludes parallelization within an instances and often becomes the bottleneck for deploying such models to latency critical scenarios. This is particularly problematic for longer texts. Here we present a convolutional architecture as an a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2022
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-031-19809-0_35